Parameter-Efficient Deep Neural Networks With Bilinear Projections

نویسندگان

چکیده

Recent research on deep neural networks (DNNs) has primarily focused improving the model accuracy. Given a proper learning framework, it is generally possible to increase depth or layer width achieve higher level of However, huge number parameters imposes more computational and memory usage overhead leads parameter redundancy. In this paper, we address redundancy problem in DNNs by replacing conventional full projections with bilinear projections. For fully-connected $D$ input nodes output nodes, applying projection can reduce space complexity from $\mathcal{O}(D^2)$ $\mathcal{O}(2D)$, achieving sub-linear size. structured lower freedom degree compared projection, causing under-fitting problem. So simply scale up mapping size increasing channels, which keep even boosts This makes very parameter-efficient handy deploy such models mobile systems limitations. Experiments four benchmark datasets show that proposed accuracies than DNNs, while significantly reduces

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

ProjectionNet: Learning Efficient On-Device Deep Networks Using Neural Projections

Deep neural networks have become ubiquitous for applications related to visual recognition and language understanding tasks. However, it is often prohibitive to use typical neural network models on devices like mobile phones or smart watches since the model sizes are huge and cannot fit in the limited memory available on such devices. While these devices could make use of machine learning model...

متن کامل

SRINIVAS, BABU: DATA-FREE PARAMETER PRUNING FOR DEEP NEURAL NETWORKS 1 Data-free Parameter Pruning for Deep Neural Networks

Deep Neural nets (NNs) with millions of parameters are at the heart of many stateof-the-art computer vision systems today. However, recent works have shown that much smaller models can achieve similar levels of performance. In this work, we address the problem of pruning parameters in a trained NN model. Instead of removing individual weights one at a time as done in previous works, we remove o...

متن کامل

Efficient Inferencing of Compressed Deep Neural Networks

Large number of weights in deep neural networks makes the models difficult to be deployed in low memory environments such as, mobile phones, IOT edge devices as well as “inferencing as a service” environments on cloud. Prior work has considered reduction in the size of the models, through compression techniques like pruning, quantization, Huffman encoding etc. However, efficient inferencing usi...

متن کامل

Efficient Model Averaging for Deep Neural Networks

Large neural networks trained on small datasets are increasingly prone to overfitting. Traditional machine learning methods can reduce overfitting by employing bagging or boosting to train several diverse models. For large neural networks, however, this is prohibitively expensive. To address this issue, we propose a method to leverage the benefits of ensembles without explicitely training sever...

متن کامل

Data-free Parameter Pruning for Deep Neural Networks

Deep Neural nets (NNs) with millions of parameters are at the heart of many stateof-the-art computer vision systems today. However, recent works have shown that much smaller models can achieve similar levels of performance. In this work, we address the problem of pruning parameters in a trained NN model. Instead of removing individual weights one at a time as done in previous works, we remove o...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: IEEE transactions on neural networks and learning systems

سال: 2021

ISSN: ['2162-237X', '2162-2388']

DOI: https://doi.org/10.1109/tnnls.2020.3016688